Learning curves from a modified VC-formalism: a case study
نویسنده
چکیده
In this paper we present a case study of a 1-dimensional higher order neuron using a statistical approach to learning theory which incorporates some information on the distribution on the sample space and can be viewed as a modification of the Vapnik-Chervonenkis formalism (VC-formalism). We concentrate on learning curves defined as averages of the worst generalization error of binary hypothesis consistent with the target on training samples as a function of the training sample size. The true learning curve is derived and compared against estimates from the classical formalism and its modification. It is shown that the modified VC-formalism improves the VC-learning curve by a factor of and also produces a meaningful result for small training sample sizes where VC-bounds are void.
منابع مشابه
Examples of learning curves from a modified VC-formalism
We examine the issue of evaluation of model specific parameters in a modified VC-formalism. Two examples are analyzed: the 2-dimensional homogeneous perceptron and the I-dimensional higher order neuron. Both models are solved theoretically, and their learning curves are compared against true learning curves. It is shown that the formalism has the potential to generate a variety of learning curv...
متن کاملMLP Can Provably Generalize Much Better than VC-bounds Indicate
Results of a study of the worst case learning curves for a particular class of probability distribution on input space to MLP with hard threshold hidden units are presented. It is shown in particular, that in the thermodynamic limit for scaling by the number of connections to the first hidden layer, although the true learning curve behaves as ~ a-I for a ~ 1, its VC-dimension based bound is tri...
متن کاملGeneralization Behaviour of Alkemic Decision Trees
This paper is concerned with generalization issues for a decision tree learner for structured data called Alkemy. Motivated by error bounds established in statistical learning theory, we study the VC dimensions of some predicate classes defined on sets and multisets – two data-modelling constructs used intensively in the knowledge representation formalism of Alkemy – and from that obtain insigh...
متن کاملEstimating Average-Case Learning Curves Using Bayesian, Statistical Physics and VC Dimension Methods
Michael Kearns· AT&T Bell Laboratories Murray Hill, New Jersey Robert Schapire AT&T Bell Laboratories Murray Hill, New Jersey In this paper we investigate an average-case model of concept learning, and give results that place the popular statistical physics and VC dimension theories of learning curve behavior in a common framework.
متن کاملTheorem 3 Let P Be a Nondegenerate Prior on F and Q Be Any Distribution on F. Let
In this paper we study a Bayesian or average-case model of concept learning with a twofold goal: to provide more precise characterizations of learning curve (sample complexity) behavior that depend on properties of both the prior distribution over concepts and the sequence of instances seen by the learner, and to smoothly unite in a common framework the popular statistical physics and VC dimens...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2005